Near-Orthogonality Regularization in Kernel Methods
نویسندگان
چکیده
Kernel methods perform nonlinear learning in high-dimensional reproducing kernel Hilbert spaces (RKHSs). Even though their large model-capacity leads to high representational power, it also incurs substantial risk of overfitting. To alleviate this problem, we propose a new regularization approach, nearorthogonality regularization, which encourages the RKHS functions to be close to being orthogonal. This effectively imposes a structural constraint over the function space, which reduces model complexity and can improve generalization performance. Besides, encouraging orthogonality reduces the redundancy among functions, which hence can reduce model size without compromising modeling power and better capture infrequent patterns in the data. Here, we define a family of orthogonality-promoting regularizers by encouraging the Gram matrix of the RKHS functions to be close to an identity matrix where the closeness is measured by Bregman matrix divergences. We apply these regularizers to two kernel methods, and develop an efficient ADMM-based algorithm to solve the regularized optimization problems. We analyze how near-orthogonality affects the generalization performance of kernel methods. Our results suggest that the closer the functions are to being orthogonal, the smaller the generalization error is. Experiments demonstrate the efficacy of near-orthogonality regularization in kernel methods.
منابع مشابه
On Approximate Birkhoff-James Orthogonality and Approximate $ast$-orthogonality in $C^ast$-algebras
We offer a new definition of $varepsilon$-orthogonality in normed spaces, and we try to explain some properties of which. Also we introduce some types of $varepsilon$-orthogonality in an arbitrary $C^ast$-algebra $mathcal{A}$, as a Hilbert $C^ast$-module over itself, and investigate some of its properties in such spaces. We state some results relating range-kernel orthogonality in $C^*$-algebras.
متن کاملSplines with non positive kernels
Non parametric regression methods can be presented in two main clusters. The one of smoothing splines methods requiring positive kernels and the other one known as Nonparametric Kernel Regression allowing the use of non positive kernels such as the Epanechnikov kernel. We propose a generalization of the smoothing spline method to include kernels which are still symmetric but not positive semi d...
متن کاملIll-Posed and Linear Inverse Problems
In this paper ill-posed linear inverse problems that arises in many applications is considered. The instability of special kind of these problems and it's relation to the kernel, is described. For finding a stable solution to these problems we need some kind of regularization that is presented. The results have been applied for a singular equation.
متن کاملPosterior Consistency of the Silverman g-prior in Bayesian Model Choice
Kernel supervised learning methods can be unified by utilizing the tools from regularization theory. The duality between regularization and prior leads to interpreting regularization methods in terms of maximum a posteriori estimation and has motivated Bayesian interpretations of kernel methods. In this paper we pursue a Bayesian interpretation of sparsity in the kernel setting by making use of...
متن کاملKernel ridge vs. principal component regression: minimax bounds and adaptability of regularization operators
Regularization is an essential element of virtually all kernel methods for nonparametric regressionproblems. A critical factor in the effectiveness of a given kernel method is the type of regularizationthat is employed. This article compares and contrasts members from a general class of regularizationtechniques, which notably includes ridge regression and principal component reg...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2017